1
Contesto del Corso e La Crisi della Riproducibilità nell'Apprendimento Automatico
EvoClass-AI002Lecture 8
00:00

Contesto del Corso e La Crisi della Riproducibilità nell'Apprendimento Automatico

Man mano che passiamo da modelli semplici e autossufficienti alle architetture complesse e a più fasi richieste per il Progetto di Milestone 1, tracciare manualmente i parametri critici su fogli di calcolo o file locali diventa completamente insostenibile. Questo flusso di lavoro complesso introduce gravi rischi per l'integrità dello sviluppo.

1. Individuazione del Punto Fermo nella Riproduzione

Il flusso di lavoro dell'apprendimento profondo prevede intrinsecamente una grande variabilità a causa di numerosi fattori (algoritmi di ottimizzazione, sottoinsiemi di dati, tecniche di regolarizzazione, differenze nell'ambiente). Senza un tracciamento sistematico, riprodurre un risultato specifico del passato – cruciale per il debug o il miglioramento di un modello in produzione – è spesso impossibile.

Cosa Deve Essere Tracciato?

Ipertparametri: All configuration settings must be recorded (e.g., Learning Rate, Batch Size, Optimizer choice, Activation function).
Stato dell'Ambiente: Software dependencies, hardware used (GPU type, OS), and exact package versions must be fixed and recorded.
Artifacts e Risultati: Pointers to the saved model weights, final metrics (Loss, Accuracy, F1 score), and training runtime must be stored.
The "Single Source of Truth" (SSOT)
Systematic experiment tracking establishes a central repository—a SSOT—where every choice made during model training is recorded automatically. This eliminates guesswork and ensures reliable auditability across all experimental runs.
conceptual_trace.py
TERMINALbash — tracking-env
> Ready. Click "Run Conceptual Trace" to see the workflow.
>
EXPERIMENT TRACE Live

Simulate the run to visualize the trace data captured.
Question 1
What is the root cause of the Deep Learning Reproducibility Crisis?
PyTorch's dependence on CUDA drivers.
The sheer number of untracked variables (code, data, hyperparameter, and environment).
The excessive memory usage of large models.
The computational cost of generating artifacts.
Question 2
In the context of MLOps, why is systematic experiment tracking essential for production?
It minimizes the total storage size of model artifacts.
It ensures that the model achieving the reported performance can be reliably reconstructed and deployed.
It speeds up the training phase of the model.
Question 3
Which element is necessary to reproduce a result but is most often forgotten in manual tracking?
The number of epochs run.
The specific versions of all Python libraries and the random seed used.
The name of the dataset used.
The time the training started.
Challenge: Tracking in Transition
Why the transition to formal tracking is non-negotiable.
You are managing 5 developers working on Milestone Project 1. Each developer reports their best model accuracy (88% to 91%) in Slack. No one can reliably tell you the exact combination of parameters or code used for the winning run.
Step 1
What immediate step must be implemented to halt the loss of critical information?
Solution:
Implement a mandatory requirement for every run to be registered with an automated tracking system before results are shared, capturing the full hyperparameter dictionary and Git hash.
Step 2
What benefit does structured tracking provide to the team that a shared spreadsheet cannot?
Solution:
Structured tracking allows automated comparison dashboards, visualizations of parameter importance, and centralized artifact storage, which is impossible with static spreadsheets.